± σ A ± t A ˆB ± σ B ± t B. (1)

Size: px
Start display at page:

Download "Â ± σ A ± t A ˆB ± σ B ± t B. (1)"

Transcription

1 Version Introduction Combining measurements which have theoretical uncertainties is a delicate matter. Indeed, we are often in the position in which sufficient information is not provided to form a completely principled combination. However, we have developed a procedure which tends to avoid the worst pitfalls. We describe this procedure here. Suppose we are given two measurements, with results expressed in the form:  ± σ A ± t A ˆB ± σ B ± t B. () Assume that  has been sampled from a probability distribution of the form p A (Â; Ā, σ A), where Ā is the mean of the distribution and σ A is the standard deviation. We make the corresponding assumption for ˆB. The t A and t B uncertainties in Eq. are the theoretical uncertainties. We may not need to know exactly what that means here, except that the same meaning should hold for both t A and t B. We suppose that both  and ˆB are measurements of the same quantity of physical interest, though possibly with quite different approaches. The question is: How do we combine our two measurements? Let the physical quantity we are trying to learn about be denoted θ. Given the two results A and B, we wish to form an estimator, ˆV for V, with statistical and theoretical uncertainties expressed separately in the form: ˆV ± σ V ± t V. () The quantities ˆV, σ V,andt V are to be computed in terms of Â, ˆB,σ A,σ B,t A, and t B. Forming the Weighted Average In the absence of theoretical uncertainties, we would normally combine our measurements according to the weighted average: ˆθ =  σ A + ˆB σ B + σ σ A B ± σ A + σ B. (3) For simplicity, we are assuming here that there is no correlation between the measurements. In general,  and ˆB will be biased estimators for V : Ā = θ + b A B = θ + b B, (4)

2 where b A and b B are the biases. We adopt the point of view that the theoretical uncertainties t A and t B are estimates related to the possible magnitudes of these biases. That is, t A b A t B b B. (5) We wish to have t V represent a similar notion. Without yet specifying the weights, assume that we continue to form ˆV as a weighted average of  and ˆB: ˆθ = w A + w B ˆB, (6) w A + w B where w A and w B are the non-negative weights. The statistical error on the weighted average is computed according to simple error propagation on the individual statistical errors: The bias for ˆθ is: σ = w A σ A + w B σ B (w A + w B ). (7) b = ˆθ θ = w Ab A + w B b B. (8) w A + w B If the theoretical uncertainties are regarded as estimates of the biases, then the theoretical uncertainty should be evaluated with the same weighting: t = w At A + w B t B, (9) w A + w B It may be noted that this posesses desirable behavior in the limit where the theoretical uncertainties are identical (completely correlated) between the two measurements: The theoretical uncertainty on ˆV is in this case the same as t A = t B ; no reduction is attained by having multiple measurements. However, it is not quite true that the theoretical uncertainties are being regarded as estimates of bias. The t A and t B provide only estimates for the magnitudes, not the signs, of the biases. Eq. 9 holds when the biases are of the same sign. If the biases are opposite sign, then we obtain t = w At A w B t B. w A + w B (0) Thus, our formula 9 breaks down in some cases. For example, suppose the theoretical uncertainties are completely anticorrelated. In the case of equal weights, the combined theoretical uncertainty should be zero, because the two uncertainties are exactly canceled in the combined result. Only a statistical uncertainty remains.

3 Unfortunately, we don t always know whether the biases are expected to have the same sign or opposite sign. As a default, we adopt the procedure of Eq. 9. In the case of similar measurements, we suspect that the sign of the bias will often have the same sign, in which case we make the right choice. In the case of quite different measurements, such as inclusive and exclusive measurements of V ub, there is no particular reason to favor either relative sign; we simply don t know. The adopted procedure has the property that it errs on the side of conservatism we will sometimes overestimate the theoretical uncertainty on the combined result. There is still a further issue. The results of the measurements themselves canprovideinformationonwhatthetheoretical uncertainty could be. Consider two measurements with negligible statistical uncertainty. Then the difference between the two measurements is the difference between the biases. If the measurements are far apart, on the scale of the theoretical uncertainties, then this is evidence that the theoretical uncertainties are of opposite sign. We make no attempt to incorporate this information, again erring on the conservative side. We turn to the question of choice of weights w A and w B. In the limit of negligible theoretical uncertainties we want to have w A = σa () w B = σb. () Using these as the weights in the presence of theoretical uncertanties can lead to undesirable behavior. For example, suppose t A t B and σ A σ B. The central value computed with only the statistical weights ignores the theoretical uncertainty. A measurement with small theoretical uncertainty may be given little weight compared to a measurement with very large theoretical uncertainty. While not wrong, this does not make optimal use of the available information. We may invent a weighting scheme which incorporates both the statistical and theoretical uncertainties, for example combining them in quadrature: w A = w B = σa + t A σb +. (3) t B Any such scheme can lead to unattractive dependence on the way measurements may be associatively combined. We nevertheless adopt this procedure, with the understanding that it is best to go back to the original measureaments when combining results, rather than making successive combinations. 3

4 3 Inconsistent Inputs It may happen that our measurements are far enough apart that they appear inconsistent in terms of the quoted uncertainties. Our primary goal in this analysis is to test consistency between available data and the standard model, including whatever theoretical uncertainties exist in the comparison. We prefer to avoid making erroneous claims of inconsistency, even at the cost of some statistical power. Thus, we presume that when two measurements of what is assumed to be the same quantity appear inconsistent, something is wrong with the measurement or with the thoeretical uncertainties in the computation. If we have no good way to determine in detail where the fault lies, we adopt a method similar to that used by the Particle Data Group (PDG) to enlarge the stated uncertainties. Given our two measurements as discussed above, we define the quantity: χ w A (Â ˆθ) + w B ( ˆB ˆθ). (4) In the limit of purely statistical and normal errors, this quantity is distributed according to a chi-square with one degree of freedom. In the more general situation here, we don t know the detailed properties, but we nonetheless use it as a measure of the consistency of the results, in the belief that the procedure we adopt will still tend to err toward conservatism. If χ, the measurements are deemed consistent. On the other hand, if χ >, we call the measurements inconsistent, and apply a scale factor to the errors in order to obtain χ =. We take the point of view that we don t know which measurement (or both) is flawed, or whether the problem is with the statistical or theoretical error evaluation. If we did have such relevant information, we could use that in a more informed procedure. Thus, we scale all of the errors (σ A, σ B, t A, t B )byafactor: S = χ. (5) This scaling does not change the central value of the averaged result, but does scale the statistical and theoretical uncertainties by the same factor. 4 Relative Errors We often are faced with the situation in which the uncertainties are relative, rather than absolute. In this case, the model in which θ is a location parameter of a Gaussian distribution breaks down. However, it may be a reasonable approximation to continue to think in terms this model, with some modification to mitigate bias. We also continue to work in the context of a least-squares minimization, although it might be interesting to investigate a maximum likelihood approach. I believe that the approach suggested here is consistent with the proposal Bob Kowalewski is making for HFAG averages. 4

5 Thus, suppose we have additional experimental uncertainties s A and s B, which scale with θ: s A = r A θ, s B = r B θ. (6) If s k is what we are given, we infer the proportionality constants according to r A = s A /Â and r B = s B / ˆB. The weights that are given in Eqn. 3 are modified to incorporate this new source of uncertainty according to: w A = σa +(r ˆθ) A + t A w B = σb +(r ˆθ). (7) B + t B Note that, as we don t know θ, weuseˆθ instead. This means that the averaging process is now iterative, until convergence to a particular value of ˆθ is obtained. Likewise, there may be a theoretical uncertainty which scales with θ, and we may treat this similarly. Thus, suppose that, for example, t A = t aa t ra, where t aa is an absolute uncertainty, and t ra = ρ A θ. We simply replace θ by ˆθ and substitute this expression wherever t A appears, e.g., in Eqn. 7. That is: w A = σa +(r ˆθ) A + t aa +. (8) ρ A ˆθ 5 Summary of Algorithm We summarize the proposed algorithm: {x i i =,,...,n} with error matrix Suppose we have n measurements M ij (x i x i )(x j x j ), (9) and mean values x i = θ + b i. (0) Note that, in the non-correlated case, M ij = σi δ ij, or including relative uncertainties, M ij = δ ij (σi + r i x i ). The parameter we are trying to learn about is θ, andtheb i is the bias that is being estimated with theoretical uncertainties t i. The present notion of the weighted average is that we find a θ which minimizes: χ = (x i θ) W ij (x j θ). () i,j This is based on the premise that we don t actually know what the biases are, and we do the minimization with zero bias in the (x θ) dependence. The 5

6 possible size of bias is taken into account in the weighting, giving more weight to those measurements in which the size of the bias is likely to be smaller. The weight matrix W in principle could be taken to be: That is, W is an estimate for (W ) ij = M ij + t i t j. () (x i θ)(x j θ) = M ij + b i b j. (3) However, we don t assume that we know the relative signs of b i and b j. Hence, the off-diagonal t i t j term in Eqn. could just as likely enter with a minus sign. We therefore use the weight matrix: (W ) ij = M ij + t i δ ij. (4) If we do know the relative signs of the biases, for example because the theoretical uncertainties are correlated, then the off-diagonal terms in Eqn. should be included, with the appropriate sign. Setting dχ /dθ θ=ˆθ = 0 gives the central value ( best estimate): i,j ˆθ = W ijx j i,j W. (5) ij The statistical uncertainty is σ = i,j (WMW) ij i,j W ij (6) Note that this reduces to σ = i,j (M ) ij, (7) in the case of only statistical uncertainites. The theoretical uncertainty is i,j t = W ijt j i,j W, (8) ij where t j t aj +(ρ j ˆθ). (9) Finally, if χ >n, these error estimates are scaled by a factor: χ S = n, (30) where χ here is the value after the minimization. 6

7 A Comparison with treating theoretical uncertainties on same footing as statistical Another approach to the present problem is to simply treat the theoretical uncertainties as if they were statistical. This procedure gives the same estimator as above for θ. However, the results for statistical and theoretical uncertainties differ in general. Let σ be the estimated statistical uncertainty on the average for this approach, and let t be the estimated theoretical uncertainty. Also, let T ij be the covariance matrix for the theoretical uncertainties in this picture. Then the statistical and theoretical uncertainties on the average are given by: σ = t = n n Note that the weights are given, as before, by i,j,k M ijw jk i,j W ij, (3) i,j,k T ijw jk i,j W. (3) ij W ij =(M + T ) ij. (33) That is, the weights are the same as the treatment earlier, if the same assumptions about theoretical correlations are made in both places. The estimates for the statistical and theoretical uncertainties differ between the two methods. That is, in general, σ σ and t t. The statistical uncertainty σ is computed from the individual statistical uncertainties according to simple error propagation. The statistical uncertainty σ is evaluated by identifying a piece of the overall quadratic combination of statistical and theoretical uncertainties as statistical. The difference between t and t is that t is computed as a weighted average of the individual t s, while t is evaluated by identifying a piece of the overall quadratic combination of statistical and theoretical uncertainties as theoretical. The approach for t is based on the notion that the theoretical uncertainties are estimates of bias, but with a conservative treatment of any unknown correlations. The t approach may be appropriate if the theoretical uncertainties are given a probablistic interpretation. Let s consider some possible special cases. Suppose that all of the t i s are the same, equal to t, and suppose that the theory uncertainties are presumed to be uncorrelated. In this case, t = t (34) t = t / n. (35) Which is more reasonable? That depends on how we view the meaning of uncorrelated in our assumption, and on whether we assign a probabilistic 7

8 interpretation to the theoretical uncertainties. If we are supposing that the acutal theoretical uncertainties are somehow randomly distributed in sign and magnitude, then it is reasonable to expect that the result will become more reliable as more numbers are averaged. However, if we consider the theoretical uncertainties as estimates of bias, which could in fact all have the same sign, then the weighted linear average is plausible. It is at least a more conservative approach in the absence of real information on the correlations. Note that if the correlation in theoretical uncertainty is actually known, the weighted linear average will take that into account. For example, suppose there are just two measurements, with t = t. If the weights are the same (that is, we also have σ = σ )thent = 0. The other approach also gives( t =0. ) σ A different illustrative case is when t =0,t 0,andM = 0 0 σ.in this case, we find ( ˆθ = ˆθ x = σ + + x )/( t σ σ + + ) t σ, (36) σ σ = (σ + + /( t ) σ σ + + ) t σ, (37) t = t σ +t + σ +t σ ( σ = σ σ + t t = t σ + t, (38) )/( + /( σ + + t σ σ + + t σ ), (39) ). (40) To understand the difference better, consider the limit in which t σ,σ : ˆθ = ˆθ σ = x σ + + x x, (4) t σ = σ, (4) σ t = t σ + 0, (43) t σ = σ, (44) t = σ. (45) In this limit, both methods agree that the important information is in x.the first method assigns a statistical error corresponding to the statistical uncertainty of x, and a theoretical uncertainty of zero, reflecting the zero theoretical uncertainty on the x measurement. The second method, however, assigns equal 8

9 statistical and theoretical uncertainites to the average. Their sum in quadrature is a plausible expression of the total uncertainty, but the breakdown into theoretical and statistical components is not reasonable. Another limit we can take in this example is σ t σ, obtaining: ˆθ = ( ˆθ t ) = x + x σ x, (46) σ = σ (47) t = t, (48) σ = t, (49) t = t. (50) Similar observations may be made in this case as in the previous. 9

arxiv: v1 [physics.data-an] 3 Jun 2008

arxiv: v1 [physics.data-an] 3 Jun 2008 arxiv:0806.0530v [physics.data-an] 3 Jun 008 Averaging Results with Theory Uncertainties F. C. Porter Lauritsen Laboratory for High Energy Physics California Institute of Technology Pasadena, California

More information

Physics 403. Segev BenZvi. Propagation of Uncertainties. Department of Physics and Astronomy University of Rochester

Physics 403. Segev BenZvi. Propagation of Uncertainties. Department of Physics and Astronomy University of Rochester Physics 403 Propagation of Uncertainties Segev BenZvi Department of Physics and Astronomy University of Rochester Table of Contents 1 Maximum Likelihood and Minimum Least Squares Uncertainty Intervals

More information

Fitting a Straight Line to Data

Fitting a Straight Line to Data Fitting a Straight Line to Data Thanks for your patience. Finally we ll take a shot at real data! The data set in question is baryonic Tully-Fisher data from http://astroweb.cwru.edu/sparc/btfr Lelli2016a.mrt,

More information

Parameter Estimation and Fitting to Data

Parameter Estimation and Fitting to Data Parameter Estimation and Fitting to Data Parameter estimation Maximum likelihood Least squares Goodness-of-fit Examples Elton S. Smith, Jefferson Lab 1 Parameter estimation Properties of estimators 3 An

More information

Gaussian Quiz. Preamble to The Humble Gaussian Distribution. David MacKay 1

Gaussian Quiz. Preamble to The Humble Gaussian Distribution. David MacKay 1 Preamble to The Humble Gaussian Distribution. David MacKay Gaussian Quiz H y y y 3. Assuming that the variables y, y, y 3 in this belief network have a joint Gaussian distribution, which of the following

More information

Physics 509: Propagating Systematic Uncertainties. Scott Oser Lecture #12

Physics 509: Propagating Systematic Uncertainties. Scott Oser Lecture #12 Physics 509: Propagating Systematic Uncertainties Scott Oser Lecture #1 1 Additive offset model Suppose we take N measurements from a distribution, and wish to estimate the true mean of the underlying

More information

Modern Methods of Data Analysis - WS 07/08

Modern Methods of Data Analysis - WS 07/08 Modern Methods of Data Analysis Lecture VIa (19.11.07) Contents: Uncertainties (II): Re: error propagation Correlated uncertainties Systematic uncertainties Re: Error Propagation (I) x = Vi,j and µi known

More information

Statistical Models with Uncertain Error Parameters (G. Cowan, arxiv: )

Statistical Models with Uncertain Error Parameters (G. Cowan, arxiv: ) Statistical Models with Uncertain Error Parameters (G. Cowan, arxiv:1809.05778) Workshop on Advanced Statistics for Physics Discovery aspd.stat.unipd.it Department of Statistical Sciences, University of

More information

EM Waves in Media. What happens when an EM wave travels through our model material?

EM Waves in Media. What happens when an EM wave travels through our model material? EM Waves in Media We can model a material as made of atoms which have a charged electron bound to a nucleus by a spring. We model the nuclei as being fixed to a grid (because they are heavy, they don t

More information

MIXED MODELS THE GENERAL MIXED MODEL

MIXED MODELS THE GENERAL MIXED MODEL MIXED MODELS This chapter introduces best linear unbiased prediction (BLUP), a general method for predicting random effects, while Chapter 27 is concerned with the estimation of variances by restricted

More information

Choosing among models

Choosing among models Eco 515 Fall 2014 Chris Sims Choosing among models September 18, 2014 c 2014 by Christopher A. Sims. This document is licensed under the Creative Commons Attribution-NonCommercial-ShareAlike 3.0 Unported

More information

Statistics and Data Analysis

Statistics and Data Analysis Statistics and Data Analysis The Crash Course Physics 226, Fall 2013 "There are three kinds of lies: lies, damned lies, and statistics. Mark Twain, allegedly after Benjamin Disraeli Statistics and Data

More information

Unobserved heterogeneity in logistic regression

Unobserved heterogeneity in logistic regression Unobserved heterogeneity in logistic regression Maarten L. Buis November 18, 2005 1 Introduction Regression is used to find the relation between various eplanatory variables and an eplained variable. For

More information

The variable θ is called the parameter of the model, and the set Ω is called the parameter space.

The variable θ is called the parameter of the model, and the set Ω is called the parameter space. Lecture 8 What is a statistical model? A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced the data. The variable θ is called

More information

CSC321 Lecture 18: Learning Probabilistic Models

CSC321 Lecture 18: Learning Probabilistic Models CSC321 Lecture 18: Learning Probabilistic Models Roger Grosse Roger Grosse CSC321 Lecture 18: Learning Probabilistic Models 1 / 25 Overview So far in this course: mainly supervised learning Language modeling

More information

A8824: Statistics Notes David Weinberg, Astronomy 8824: Statistics Notes 6 Estimating Errors From Data

A8824: Statistics Notes David Weinberg, Astronomy 8824: Statistics Notes 6 Estimating Errors From Data Where does the error bar go? Astronomy 8824: Statistics otes 6 Estimating Errors From Data Suppose you measure the average depression of flux in a quasar caused by absorption from the Lyman-alpha forest.

More information

MS&E 226: Small Data

MS&E 226: Small Data MS&E 226: Small Data Lecture 12: Frequentist properties of estimators (v4) Ramesh Johari ramesh.johari@stanford.edu 1 / 39 Frequentist inference 2 / 39 Thinking like a frequentist Suppose that for some

More information

Linear Regression Models

Linear Regression Models Linear Regression Models Model Description and Model Parameters Modelling is a central theme in these notes. The idea is to develop and continuously improve a library of predictive models for hazards,

More information

So, what are special sciences? ones that are particularly dear to the author? ( Oh dear. I am touched. Psychology is just, so, well, special!

So, what are special sciences? ones that are particularly dear to the author? ( Oh dear. I am touched. Psychology is just, so, well, special! Jerry Fodor and his Special Sciences So, what are special sciences? ones that are particularly dear to the author? ( Oh dear. I am touched. Psychology is just, so, well, special! ) The use of special in

More information

Purposes of Data Analysis. Variables and Samples. Parameters and Statistics. Part 1: Probability Distributions

Purposes of Data Analysis. Variables and Samples. Parameters and Statistics. Part 1: Probability Distributions Part 1: Probability Distributions Purposes of Data Analysis True Distributions or Relationships in the Earths System Probability Distribution Normal Distribution Student-t Distribution Chi Square Distribution

More information

Parameter Uncertainties in IMFIT CASA Requirements Document Brian S. Mason (NRAO) March 5, 2014

Parameter Uncertainties in IMFIT CASA Requirements Document Brian S. Mason (NRAO) March 5, 2014 Parameter Uncertainties in IMFIT CASA Requirements Document Brian S. Mason (NRAO) March 5, 014 The CASA task IMFIT fits elliptical Gaussians to image data. As reported by several users in CASA tickets

More information

Frequentist-Bayesian Model Comparisons: A Simple Example

Frequentist-Bayesian Model Comparisons: A Simple Example Frequentist-Bayesian Model Comparisons: A Simple Example Consider data that consist of a signal y with additive noise: Data vector (N elements): D = y + n The additive noise n has zero mean and diagonal

More information

Hypothesis testing. Chapter Formulating a hypothesis. 7.2 Testing if the hypothesis agrees with data

Hypothesis testing. Chapter Formulating a hypothesis. 7.2 Testing if the hypothesis agrees with data Chapter 7 Hypothesis testing 7.1 Formulating a hypothesis Up until now we have discussed how to define a measurement in terms of a central value, uncertainties, and units, as well as how to extend these

More information

Histograms were invented for human beings, so that we can visualize the observations as a density (estimated values of a pdf).

Histograms were invented for human beings, so that we can visualize the observations as a density (estimated values of a pdf). How to fit a histogram The Histogram Histograms were invented for human beings, so that we can visualize the observations as a density (estimated values of a pdf). The binning most convenient for visualization

More information

4th IMPRS Astronomy Summer School Drawing Astrophysical Inferences from Data Sets

4th IMPRS Astronomy Summer School Drawing Astrophysical Inferences from Data Sets 4th IMPRS Astronomy Summer School Drawing Astrophysical Inferences from Data Sets William H. Press The University of Texas at Austin Lecture 6 IMPRS Summer School 2009, Prof. William H. Press 1 Mixture

More information

PHYSTAT2003, SLAC, Stanford, California, September 8-11, 2003

PHYSTAT2003, SLAC, Stanford, California, September 8-11, 2003 Asymmetric Errors PHYSTAT003, SLAC, Stanford, California, September 8-11, 003 Roger Barlow Manchester University, UK and Stanford University, USA Errors quoted on results are often given in asymmetric

More information

Systematic uncertainties in statistical data analysis for particle physics. DESY Seminar Hamburg, 31 March, 2009

Systematic uncertainties in statistical data analysis for particle physics. DESY Seminar Hamburg, 31 March, 2009 Systematic uncertainties in statistical data analysis for particle physics DESY Seminar Hamburg, 31 March, 2009 Glen Cowan Physics Department Royal Holloway, University of London g.cowan@rhul.ac.uk www.pp.rhul.ac.uk/~cowan

More information

DS-GA 1002 Lecture notes 11 Fall Bayesian statistics

DS-GA 1002 Lecture notes 11 Fall Bayesian statistics DS-GA 100 Lecture notes 11 Fall 016 Bayesian statistics In the frequentist paradigm we model the data as realizations from a distribution that depends on deterministic parameters. In contrast, in Bayesian

More information

SIO 221B, Rudnick adapted from Davis 1. 1 x lim. N x 2 n = 1 N. { x} 1 N. N x = 1 N. N x = 1 ( N N x ) x = 0 (3) = 1 x N 2

SIO 221B, Rudnick adapted from Davis 1. 1 x lim. N x 2 n = 1 N. { x} 1 N. N x = 1 N. N x = 1 ( N N x ) x = 0 (3) = 1 x N 2 SIO B, Rudnick adapted from Davis VII. Sampling errors We do not have access to the true statistics, so we must compute sample statistics. By this we mean that the number of realizations we average over

More information

Treatment of Error in Experimental Measurements

Treatment of Error in Experimental Measurements in Experimental Measurements All measurements contain error. An experiment is truly incomplete without an evaluation of the amount of error in the results. In this course, you will learn to use some common

More information

Bayesian model selection: methodology, computation and applications

Bayesian model selection: methodology, computation and applications Bayesian model selection: methodology, computation and applications David Nott Department of Statistics and Applied Probability National University of Singapore Statistical Genomics Summer School Program

More information

Journeys of an Accidental Statistician

Journeys of an Accidental Statistician Journeys of an Accidental Statistician A partially anecdotal account of A Unified Approach to the Classical Statistical Analysis of Small Signals, GJF and Robert D. Cousins, Phys. Rev. D 57, 3873 (1998)

More information

6.867 Machine Learning

6.867 Machine Learning 6.867 Machine Learning Problem set 1 Solutions Thursday, September 19 What and how to turn in? Turn in short written answers to the questions explicitly stated, and when requested to explain or prove.

More information

DS-GA 1003: Machine Learning and Computational Statistics Homework 7: Bayesian Modeling

DS-GA 1003: Machine Learning and Computational Statistics Homework 7: Bayesian Modeling DS-GA 1003: Machine Learning and Computational Statistics Homework 7: Bayesian Modeling Due: Tuesday, May 10, 2016, at 6pm (Submit via NYU Classes) Instructions: Your answers to the questions below, including

More information

arxiv:hep-ex/ v1 2 Jun 2000

arxiv:hep-ex/ v1 2 Jun 2000 MPI H - V7-000 May 3, 000 Averaging Measurements with Hidden Correlations and Asymmetric Errors Michael Schmelling / MPI for Nuclear Physics Postfach 03980, D-6909 Heidelberg arxiv:hep-ex/0006004v Jun

More information

The Multivariate Gaussian Distribution [DRAFT]

The Multivariate Gaussian Distribution [DRAFT] The Multivariate Gaussian Distribution DRAFT David S. Rosenberg Abstract This is a collection of a few key and standard results about multivariate Gaussian distributions. I have not included many proofs,

More information

Econ 836 Final Exam. 2 w N 2 u N 2. 2 v N

Econ 836 Final Exam. 2 w N 2 u N 2. 2 v N 1) [4 points] Let Econ 836 Final Exam Y Xβ+ ε, X w+ u, w N w~ N(, σi ), u N u~ N(, σi ), ε N ε~ Nu ( γσ, I ), where X is a just one column. Let denote the OLS estimator, and define residuals e as e Y X.

More information

Vectors Part 1: Two Dimensions

Vectors Part 1: Two Dimensions Vectors Part 1: Two Dimensions Last modified: 20/02/2018 Links Scalars Vectors Definition Notation Polar Form Compass Directions Basic Vector Maths Multiply a Vector by a Scalar Unit Vectors Example Vectors

More information

CSC 578 Neural Networks and Deep Learning

CSC 578 Neural Networks and Deep Learning CSC 578 Neural Networks and Deep Learning Fall 2018/19 3. Improving Neural Networks (Some figures adapted from NNDL book) 1 Various Approaches to Improve Neural Networks 1. Cost functions Quadratic Cross

More information

Overview. Probabilistic Interpretation of Linear Regression Maximum Likelihood Estimation Bayesian Estimation MAP Estimation

Overview. Probabilistic Interpretation of Linear Regression Maximum Likelihood Estimation Bayesian Estimation MAP Estimation Overview Probabilistic Interpretation of Linear Regression Maximum Likelihood Estimation Bayesian Estimation MAP Estimation Probabilistic Interpretation: Linear Regression Assume output y is generated

More information

1 Motivation for Instrumental Variable (IV) Regression

1 Motivation for Instrumental Variable (IV) Regression ECON 370: IV & 2SLS 1 Instrumental Variables Estimation and Two Stage Least Squares Econometric Methods, ECON 370 Let s get back to the thiking in terms of cross sectional (or pooled cross sectional) data

More information

MACHINE LEARNING INTRODUCTION: STRING CLASSIFICATION

MACHINE LEARNING INTRODUCTION: STRING CLASSIFICATION MACHINE LEARNING INTRODUCTION: STRING CLASSIFICATION THOMAS MAILUND Machine learning means different things to different people, and there is no general agreed upon core set of algorithms that must be

More information

Interval estimation. October 3, Basic ideas CLT and CI CI for a population mean CI for a population proportion CI for a Normal mean

Interval estimation. October 3, Basic ideas CLT and CI CI for a population mean CI for a population proportion CI for a Normal mean Interval estimation October 3, 2018 STAT 151 Class 7 Slide 1 Pandemic data Treatment outcome, X, from n = 100 patients in a pandemic: 1 = recovered and 0 = not recovered 1 1 1 0 0 0 1 1 1 0 0 1 0 1 0 0

More information

On the errors introduced by the naive Bayes independence assumption

On the errors introduced by the naive Bayes independence assumption On the errors introduced by the naive Bayes independence assumption Author Matthijs de Wachter 3671100 Utrecht University Master Thesis Artificial Intelligence Supervisor Dr. Silja Renooij Department of

More information

CSC 2541: Bayesian Methods for Machine Learning

CSC 2541: Bayesian Methods for Machine Learning CSC 2541: Bayesian Methods for Machine Learning Radford M. Neal, University of Toronto, 2011 Lecture 10 Alternatives to Monte Carlo Computation Since about 1990, Markov chain Monte Carlo has been the dominant

More information

ECONOMICS 210C / ECONOMICS 236A MONETARY HISTORY

ECONOMICS 210C / ECONOMICS 236A MONETARY HISTORY Fall 018 University of California, Berkeley Christina Romer David Romer ECONOMICS 10C / ECONOMICS 36A MONETARY HISTORY A Little about GLS, Heteroskedasticity-Consistent Standard Errors, Clustering, and

More information

BAYESIAN ESTIMATION OF LINEAR STATISTICAL MODEL BIAS

BAYESIAN ESTIMATION OF LINEAR STATISTICAL MODEL BIAS BAYESIAN ESTIMATION OF LINEAR STATISTICAL MODEL BIAS Andrew A. Neath 1 and Joseph E. Cavanaugh 1 Department of Mathematics and Statistics, Southern Illinois University, Edwardsville, Illinois 606, USA

More information

Probabilistic Graphical Models

Probabilistic Graphical Models 2016 Robert Nowak Probabilistic Graphical Models 1 Introduction We have focused mainly on linear models for signals, in particular the subspace model x = Uθ, where U is a n k matrix and θ R k is a vector

More information

Ph 219/CS 219. Exercises Due: Friday 20 October 2006

Ph 219/CS 219. Exercises Due: Friday 20 October 2006 1 Ph 219/CS 219 Exercises Due: Friday 20 October 2006 1.1 How far apart are two quantum states? Consider two quantum states described by density operators ρ and ρ in an N-dimensional Hilbert space, and

More information

Measurement And Uncertainty

Measurement And Uncertainty Measurement And Uncertainty Based on Guidelines for Evaluating and Expressing the Uncertainty of NIST Measurement Results, NIST Technical Note 1297, 1994 Edition PHYS 407 1 Measurement approximates or

More information

Wooldridge, Introductory Econometrics, 4th ed. Chapter 15: Instrumental variables and two stage least squares

Wooldridge, Introductory Econometrics, 4th ed. Chapter 15: Instrumental variables and two stage least squares Wooldridge, Introductory Econometrics, 4th ed. Chapter 15: Instrumental variables and two stage least squares Many economic models involve endogeneity: that is, a theoretical relationship does not fit

More information

IV Estimation and its Limitations: Weak Instruments and Weakly Endogeneous Regressors

IV Estimation and its Limitations: Weak Instruments and Weakly Endogeneous Regressors IV Estimation and its Limitations: Weak Instruments and Weakly Endogeneous Regressors Laura Mayoral IAE, Barcelona GSE and University of Gothenburg Gothenburg, May 2015 Roadmap Deviations from the standard

More information

ECON 4160, Spring term 2015 Lecture 7

ECON 4160, Spring term 2015 Lecture 7 ECON 4160, Spring term 2015 Lecture 7 Identification and estimation of SEMs (Part 1) Ragnar Nymoen Department of Economics 8 Oct 2015 1 / 55 HN Ch 15 References to Davidson and MacKinnon, Ch 8.1-8.5 Ch

More information

Introductory Econometrics. Review of statistics (Part II: Inference)

Introductory Econometrics. Review of statistics (Part II: Inference) Introductory Econometrics Review of statistics (Part II: Inference) Jun Ma School of Economics Renmin University of China October 1, 2018 1/16 Null and alternative hypotheses Usually, we have two competing

More information

Statistics. Lecture 4 August 9, 2000 Frank Porter Caltech. 1. The Fundamentals; Point Estimation. 2. Maximum Likelihood, Least Squares and All That

Statistics. Lecture 4 August 9, 2000 Frank Porter Caltech. 1. The Fundamentals; Point Estimation. 2. Maximum Likelihood, Least Squares and All That Statistics Lecture 4 August 9, 2000 Frank Porter Caltech The plan for these lectures: 1. The Fundamentals; Point Estimation 2. Maximum Likelihood, Least Squares and All That 3. What is a Confidence Interval?

More information

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3 Hypothesis Testing CB: chapter 8; section 0.3 Hypothesis: statement about an unknown population parameter Examples: The average age of males in Sweden is 7. (statement about population mean) The lowest

More information

INTERPENETRATING SAMPLES

INTERPENETRATING SAMPLES INTERPENETRATING SAMPLES Dr. A. C. Kulshreshtha UN Statistical Institute for Asia and Pacific Second RAP Regional Workshop on Building Training Resources for Improving Agricultural & Rural Statistics Sampling

More information

Bayesian vs frequentist techniques for the analysis of binary outcome data

Bayesian vs frequentist techniques for the analysis of binary outcome data 1 Bayesian vs frequentist techniques for the analysis of binary outcome data By M. Stapleton Abstract We compare Bayesian and frequentist techniques for analysing binary outcome data. Such data are commonly

More information

Review of Maximum Likelihood Estimators

Review of Maximum Likelihood Estimators Libby MacKinnon CSE 527 notes Lecture 7, October 7, 2007 MLE and EM Review of Maximum Likelihood Estimators MLE is one of many approaches to parameter estimation. The likelihood of independent observations

More information

An Empirical-Bayes Score for Discrete Bayesian Networks

An Empirical-Bayes Score for Discrete Bayesian Networks An Empirical-Bayes Score for Discrete Bayesian Networks scutari@stats.ox.ac.uk Department of Statistics September 8, 2016 Bayesian Network Structure Learning Learning a BN B = (G, Θ) from a data set D

More information

32. STATISTICS. 32. Statistics 1

32. STATISTICS. 32. Statistics 1 32. STATISTICS 32. Statistics 1 Revised September 2007 by G. Cowan (RHUL). This chapter gives an overview of statistical methods used in High Energy Physics. In statistics, we are interested in using a

More information

1 What does the random effect η mean?

1 What does the random effect η mean? Some thoughts on Hanks et al, Environmetrics, 2015, pp. 243-254. Jim Hodges Division of Biostatistics, University of Minnesota, Minneapolis, Minnesota USA 55414 email: hodge003@umn.edu October 13, 2015

More information

Statistics: Learning models from data

Statistics: Learning models from data DS-GA 1002 Lecture notes 5 October 19, 2015 Statistics: Learning models from data Learning models from data that are assumed to be generated probabilistically from a certain unknown distribution is a crucial

More information

STATS 200: Introduction to Statistical Inference. Lecture 29: Course review

STATS 200: Introduction to Statistical Inference. Lecture 29: Course review STATS 200: Introduction to Statistical Inference Lecture 29: Course review Course review We started in Lecture 1 with a fundamental assumption: Data is a realization of a random process. The goal throughout

More information

Uncertainty and Rules

Uncertainty and Rules Uncertainty and Rules We have already seen that expert systems can operate within the realm of uncertainty. There are several sources of uncertainty in rules: Uncertainty related to individual rules Uncertainty

More information

Introduction to Bayesian Learning. Machine Learning Fall 2018

Introduction to Bayesian Learning. Machine Learning Fall 2018 Introduction to Bayesian Learning Machine Learning Fall 2018 1 What we have seen so far What does it mean to learn? Mistake-driven learning Learning by counting (and bounding) number of mistakes PAC learnability

More information

Detection and quantification capabilities

Detection and quantification capabilities 18.4.3.7 Detection and quantification capabilities Among the most important Performance Characteristics of the Chemical Measurement Process (CMP) are those that can serve as measures of the underlying

More information

Physics 403. Segev BenZvi. Parameter Estimation, Correlations, and Error Bars. Department of Physics and Astronomy University of Rochester

Physics 403. Segev BenZvi. Parameter Estimation, Correlations, and Error Bars. Department of Physics and Astronomy University of Rochester Physics 403 Parameter Estimation, Correlations, and Error Bars Segev BenZvi Department of Physics and Astronomy University of Rochester Table of Contents 1 Review of Last Class Best Estimates and Reliability

More information

Bayesian inference: what it means and why we care

Bayesian inference: what it means and why we care Bayesian inference: what it means and why we care Robin J. Ryder Centre de Recherche en Mathématiques de la Décision Université Paris-Dauphine 6 November 2017 Mathematical Coffees Robin Ryder (Dauphine)

More information

Class President: A Network Approach to Popularity. Due July 18, 2014

Class President: A Network Approach to Popularity. Due July 18, 2014 Class President: A Network Approach to Popularity Due July 8, 24 Instructions. Due Fri, July 8 at :59 PM 2. Work in groups of up to 3 3. Type up the report, and submit as a pdf on D2L 4. Attach the code

More information

Introduction: MLE, MAP, Bayesian reasoning (28/8/13)

Introduction: MLE, MAP, Bayesian reasoning (28/8/13) STA561: Probabilistic machine learning Introduction: MLE, MAP, Bayesian reasoning (28/8/13) Lecturer: Barbara Engelhardt Scribes: K. Ulrich, J. Subramanian, N. Raval, J. O Hollaren 1 Classifiers In this

More information

37-6 Watching the electrons (matter waves)

37-6 Watching the electrons (matter waves) 37-6 Watching the electrons (matter waves) 1 testing our proposition: the electrons go either through hole 1 or hole 2 add a very strong light source behind walls between two holes, electrons will scatter

More information

Strati cation in Multivariate Modeling

Strati cation in Multivariate Modeling Strati cation in Multivariate Modeling Tihomir Asparouhov Muthen & Muthen Mplus Web Notes: No. 9 Version 2, December 16, 2004 1 The author is thankful to Bengt Muthen for his guidance, to Linda Muthen

More information

Quantitative Introduction ro Risk and Uncertainty in Business Module 5: Hypothesis Testing

Quantitative Introduction ro Risk and Uncertainty in Business Module 5: Hypothesis Testing Quantitative Introduction ro Risk and Uncertainty in Business Module 5: Hypothesis Testing M. Vidyasagar Cecil & Ida Green Chair The University of Texas at Dallas Email: M.Vidyasagar@utdallas.edu October

More information

Measurement Error PHYS Introduction

Measurement Error PHYS Introduction PHYS 1301 Measurement Error Introduction We have confidence that a particular physics theory is telling us something interesting about the physical universe because we are able to test quantitatively its

More information

Week 3: Linear Regression

Week 3: Linear Regression Week 3: Linear Regression Instructor: Sergey Levine Recap In the previous lecture we saw how linear regression can solve the following problem: given a dataset D = {(x, y ),..., (x N, y N )}, learn to

More information

Discrete Mathematics and Probability Theory Spring 2014 Anant Sahai Note 10

Discrete Mathematics and Probability Theory Spring 2014 Anant Sahai Note 10 EECS 70 Discrete Mathematics and Probability Theory Spring 2014 Anant Sahai Note 10 Introduction to Basic Discrete Probability In the last note we considered the probabilistic experiment where we flipped

More information

Unified approach to the classical statistical analysis of small signals

Unified approach to the classical statistical analysis of small signals PHYSICAL REVIEW D VOLUME 57, NUMBER 7 1 APRIL 1998 Unified approach to the classical statistical analysis of small signals Gary J. Feldman * Department of Physics, Harvard University, Cambridge, Massachusetts

More information

UC Berkeley Department of Electrical Engineering and Computer Science Department of Statistics. EECS 281A / STAT 241A Statistical Learning Theory

UC Berkeley Department of Electrical Engineering and Computer Science Department of Statistics. EECS 281A / STAT 241A Statistical Learning Theory UC Berkeley Department of Electrical Engineering and Computer Science Department of Statistics EECS 281A / STAT 241A Statistical Learning Theory Solutions to Problem Set 2 Fall 2011 Issued: Wednesday,

More information

DEMBSKI S SPECIFIED COMPLEXITY: A SIMPLE ERROR IN ARITHMETIC

DEMBSKI S SPECIFIED COMPLEXITY: A SIMPLE ERROR IN ARITHMETIC DEMBSKI S SPECIFIED COMPLEXITY: A SIMPLE ERROR IN ARITHMETIC HOWARD A. LANDMAN Abstract. We show that the derivation of Dembski s formula for specified complexity contains a simple but enormous error,

More information

Measurement Error PHYS Introduction

Measurement Error PHYS Introduction PHYS 1301 Measurement Error Introduction We have confidence that a particular physics theory is telling us something interesting about the physical universe because we are able to test quantitatively its

More information

Frank Porter February 26, 2013

Frank Porter February 26, 2013 116 Frank Porter February 26, 2013 Chapter 6 Hypothesis Tests Often, we want to address questions such as whether the possible observation of a new effect is really significant, or merely a chance fluctuation.

More information

Advanced Statistical Methods. Lecture 6

Advanced Statistical Methods. Lecture 6 Advanced Statistical Methods Lecture 6 Convergence distribution of M.-H. MCMC We denote the PDF estimated by the MCMC as. It has the property Convergence distribution After some time, the distribution

More information

Continuum Probability and Sets of Measure Zero

Continuum Probability and Sets of Measure Zero Chapter 3 Continuum Probability and Sets of Measure Zero In this chapter, we provide a motivation for using measure theory as a foundation for probability. It uses the example of random coin tossing to

More information

Lecture 3: Linear Models. Bruce Walsh lecture notes Uppsala EQG course version 28 Jan 2012

Lecture 3: Linear Models. Bruce Walsh lecture notes Uppsala EQG course version 28 Jan 2012 Lecture 3: Linear Models Bruce Walsh lecture notes Uppsala EQG course version 28 Jan 2012 1 Quick Review of the Major Points The general linear model can be written as y = X! + e y = vector of observed

More information

Ages of stellar populations from color-magnitude diagrams. Paul Baines. September 30, 2008

Ages of stellar populations from color-magnitude diagrams. Paul Baines. September 30, 2008 Ages of stellar populations from color-magnitude diagrams Paul Baines Department of Statistics Harvard University September 30, 2008 Context & Example Welcome! Today we will look at using hierarchical

More information

A Probability Review

A Probability Review A Probability Review Outline: A probability review Shorthand notation: RV stands for random variable EE 527, Detection and Estimation Theory, # 0b 1 A Probability Review Reading: Go over handouts 2 5 in

More information

ECON 4160, Autumn term Lecture 1

ECON 4160, Autumn term Lecture 1 ECON 4160, Autumn term 2017. Lecture 1 a) Maximum Likelihood based inference. b) The bivariate normal model Ragnar Nymoen University of Oslo 24 August 2017 1 / 54 Principles of inference I Ordinary least

More information

An improved procedure for combining Type A and Type B components of measurement uncertainty

An improved procedure for combining Type A and Type B components of measurement uncertainty Int. J. Metrol. Qual. Eng. 4, 55 62 (2013) c EDP Sciences 2013 DOI: 10.1051/ijmqe/2012038 An improved procedure for combining Type A and Type B components of measurement uncertainty R. Willink Received:

More information

For more information about how to cite these materials visit

For more information about how to cite these materials visit Author(s): Kerby Shedden, Ph.D., 2010 License: Unless otherwise noted, this material is made available under the terms of the Creative Commons Attribution Share Alike 3.0 License: http://creativecommons.org/licenses/by-sa/3.0/

More information

Lecture 24: Weighted and Generalized Least Squares

Lecture 24: Weighted and Generalized Least Squares Lecture 24: Weighted and Generalized Least Squares 1 Weighted Least Squares When we use ordinary least squares to estimate linear regression, we minimize the mean squared error: MSE(b) = 1 n (Y i X i β)

More information

Quadrature for the Finite Free Convolution

Quadrature for the Finite Free Convolution Spectral Graph Theory Lecture 23 Quadrature for the Finite Free Convolution Daniel A. Spielman November 30, 205 Disclaimer These notes are not necessarily an accurate representation of what happened in

More information

Homework #2 Due Monday, April 18, 2012

Homework #2 Due Monday, April 18, 2012 12.540 Homework #2 Due Monday, April 18, 2012 Matlab solution codes are given in HW02_2012.m This code uses cells and contains the solutions to all the questions. Question 1: Non-linear estimation problem

More information

Brandon C. Kelly (Harvard Smithsonian Center for Astrophysics)

Brandon C. Kelly (Harvard Smithsonian Center for Astrophysics) Brandon C. Kelly (Harvard Smithsonian Center for Astrophysics) Probability quantifies randomness and uncertainty How do I estimate the normalization and logarithmic slope of a X ray continuum, assuming

More information

Simple groups and the classification of finite groups

Simple groups and the classification of finite groups Simple groups and the classification of finite groups 1 Finite groups of small order How can we describe all finite groups? Before we address this question, let s write down a list of all the finite groups

More information

A Bayesian Treatment of Linear Gaussian Regression

A Bayesian Treatment of Linear Gaussian Regression A Bayesian Treatment of Linear Gaussian Regression Frank Wood December 3, 2009 Bayesian Approach to Classical Linear Regression In classical linear regression we have the following model y β, σ 2, X N(Xβ,

More information

32. STATISTICS. 32. Statistics 1

32. STATISTICS. 32. Statistics 1 32. STATISTICS 32. Statistics 1 Revised April 1998 by F. James (CERN); February 2000 by R. Cousins (UCLA); October 2001, October 2003, and August 2005 by G. Cowan (RHUL). This chapter gives an overview

More information

Chapter 7: Hypothesis testing

Chapter 7: Hypothesis testing Chapter 7: Hypothesis testing Hypothesis testing is typically done based on the cumulative hazard function. Here we ll use the Nelson-Aalen estimate of the cumulative hazard. The survival function is used

More information

Handout 1 Systems of linear equations Gaussian elimination

Handout 1 Systems of linear equations Gaussian elimination 06-21254 Mathematical Techniques for Computer Science The University of Birmingham Autumn Semester 2017 School of Computer Science c Achim Jung September 27, 2017 Handout 1 Systems of linear equations

More information

Testing Restrictions and Comparing Models

Testing Restrictions and Comparing Models Econ. 513, Time Series Econometrics Fall 00 Chris Sims Testing Restrictions and Comparing Models 1. THE PROBLEM We consider here the problem of comparing two parametric models for the data X, defined by

More information